142 research outputs found
ELSIM: End-to-end learning of reusable skills through intrinsic motivation
Taking inspiration from developmental learning, we present a novel
reinforcement learning architecture which hierarchically learns and represents
self-generated skills in an end-to-end way. With this architecture, an agent
focuses only on task-rewarded skills while keeping the learning process of
skills bottom-up. This bottom-up approach allows to learn skills that 1- are
transferable across tasks, 2- improves exploration when rewards are sparse. To
do so, we combine a previously defined mutual information objective with a
novel curriculum learning algorithm, creating an unlimited and explorable tree
of skills. We test our agent on simple gridworld environments to understand and
visualize how the agent distinguishes between its skills. Then we show that our
approach can scale on more difficult MuJoCo environments in which our agent is
able to build a representation of skills which improve over a baseline both
transfer learning and exploration when rewards are sparse.Comment: Accepted at ECML 202
Proceedings of The Multi-Agent Logics, Languages, and Organisations Federated Workshops (MALLOW 2010)
http://ceur-ws.org/Vol-627/allproceedings.pdfInternational audienceMALLOW-2010 is a third edition of a series initiated in 2007 in Durham, and pursued in 2009 in Turin. The objective, as initially stated, is to "provide a venue where: the cost of participation was minimum; participants were able to attend various workshops, so fostering collaboration and cross-fertilization; there was a friendly atmosphere and plenty of time for networking, by maximizing the time participants spent together"
Étude de la motivation intrinsèque en apprentissage par renforcement
National audienceDespite many existing works in reinforcement learning (RL) and the recent successes obtained by combining it with deep learning, RL is facing many challenges. Some of them, like the ability to abstract the action or the difficulty to conceive a reward function without expert knowledge, can be addressed by the use of intrinsic motivation. In this article, we provide a survey on the role of intrinsic motivation in RL and its different usages by detailing interests and limits of existing approaches. Our analysis suggests that mutual information is central to most of the work using intrinsic motivation in RL. The combination of deep RL and intrinsic motivation enables to learn more complicated and more generalisable behaviours than what enables standard RL.Malgré les nombreux travaux existants en apprentissage par renforcement (AR) et les récents succès obtenus notamment en le combinant avec l'apprentissage profond, l'AR fait encore aujourd'hui face à de nombreux défis. Certains d'entre eux, comme la problématique de l'abstraction temporelle des actions ou la difficulté de concevoir une fonction de récompense sans connaissances ex-pertes, peuvent être adressées par l'utilisation de récompenses intrinsèques. Dans cet article, nous proposons une étude du rôle de la motivation intrinsèque en AR et de ses différents usages, en détaillant les intérêts et les limites des approches existantes. Notre analyse suggère que la notion d'information mutuelle est centrale à la plupart des travaux utilisant la motivation intrinsèque en AR. Celle-ci, combinée aux algorithmes d'AR profond, permet d'apprendre des comportements plus complexes et plus généralisables que ce que permet l'AR traditionnel
Developmental Learning for Social Robots in Real-World Interactions
International audienceThis paper reports preliminary research work on applying developmental learning to social robotics for making human-robot interactions more instinctive and more natural. Developmental learning is an unsupervised learning strategy relying on the fact that the learning agent is intrinsically motivated, and is able to incrementally build its own representation of the world through its experiences of interaction with it. Our claim is that using developmental learning in social robots could dramatically change the way we envision human-robot interaction, notably by giving the robot an active role in the interaction building process, and even more importantly, in the way it autonomously learns suitable behaviors over time. Developmental learning appears to be an appropriate approach to develop a form of "interactional intelligence" for social robots. In this work, our goal was to set up a common framework for implementing, experimenting and evaluating developmental learning algorithms with various social robots
Sequential annotations for naturally-occurring HRI: first insights
We explain the methodology we developed for improving the interactions
accomplished by an embedded conversational agent, drawing from Conversation
Analytic sequential and multimodal analysis. The use case is a Pepper robot
that is expected to inform and orient users in a library. In order to propose
and learn better interactive schema, we are creating a corpus of
naturally-occurring interactions that will be made available to the community.
To do so, we propose an annotation practice based on some theoretical
underpinnings about the use of language and multimodal resources in human-robot
interaction. CCS CONCEPTS Computing methodologies
Discourse, dialogue and pragmatics; Human-centered computing
Text input; HCI theory, concepts and models; Field studies.Comment: Peer-reviewed workshop paper accepted for the ''Human-Robot
Conversational Interaction'' workshop that took place at the ''ACM/IEEE
International Conference on Human-Robot Interaction'' 2023 Conference in
Stockholm, Swede
Using swarm intelligence for dynamic web content organization
International audienceNot availableNon disponibl
Programming complex systems with the multiagents paradigm (Invited tutoriel)
International audienc
- …